Language Models
12 Threads
Permit me to pique your interest: Self-Taught Optimizer (STOP) This paper reveals a powerful new capability of large language models - the ability to recursively improve how they...
Do language models have an internal world model? A sense of time? At multiple spatiotemporal scales? In a new paper with @tegmark we provide evidence that they do by finding a lit...
list of some of the most popular LLMs and LLMs datasets: LLMs Bard (Google AI) Bard large language model ChatGPT (OpenAI) ChatGPT large language model Claude (Anthropic) Claude...
5 Advanced ChatGPT prompt techniques that will put you ahead of the world: (🧵 A thread)
Nearly all recently-proposed large language models (LLMs) are based upon the decoder-only transformer architecture. But, is this always the best architecture to use? It depends… 🧵...
there are lots of threads like “THE 10 best prompts for ChatGPT” this is not one of those prompt engineering is evolving beyond simple ideas like few-shot learning and CoT reason...
Each “block” of a large language model (LLM) is comprised of self-attention and a feed-forward transformation. However, the exact self-attention variant used by LLMs is masked, mul...
ChatGPT is all the rave. But what does GPT actually mean? Here’s a quick breakdown so you don’t get left behind:
1/ In 2021, we shared next-gen language + conversation capabilities powered by our Language Model for Dialogue Applications (LaMDA). Coming soon: Bard, a new experimental conversat...
Three years in the making - our big review/position piece on the capabilities of large language models (LLMs) from the cognitive science perspective. Thread below! 1/ https://t....
In text generation, how do you get a large language model to be more (or less) creative? 🎨 Depending on your use case, you may want the model to be: 1. Very creative, or 2. Very p...
@schachin @cemper ? They are based on language models, derived from corpuses of content. Things like Wiki (factual), news (factual/perspective) etc. Break away from those, and the...